Progressive Ladder Networks for Semi-Supervised Transfer Learning
نویسنده
چکیده
Semi-supervised learning has achieved remarkable success in the past few years at harnessing the power of unlabeled data and tackling domains where few labeled data examples exist. We test the hypothesis that deep semisupervised architectures learn general representations. We combine two well-known techniques for semi-supervised and transfer learning, ladder networks and progressive neural networks, to create the progressive ladder network, a framework for transferring from semi-supervised representations to supervised representations. We use this framework for domain adaptation in the digit recognition domain, from the MNIST handwritten digits dataset to the Street View House Numbers (SVHN) dataset. Our experiments show that semi-supervised representations are more transferable than supervised learning representations. We find that using our progressive ladder network architecture on simple fully connected neural networks trained on SVHN can yield an increase in test accuracy of about 40%, from around 35% to 77.7%.
منابع مشابه
Semi-supervised Phoneme Recognition with Recurrent Ladder Networks
Ladder networks are a notable new concept in the field of semi-supervised learning by showing state-of-the-art results in image recognition tasks while being compatible with many existing neural architectures. We present the recurrent ladder network, a novel modification of the ladder network, for semi-supervised learning of recurrent neural networks which we evaluate with a phoneme recognition...
متن کاملSemi-supervised Learning with Ladder Networks
We combine supervised learning with unsupervised learning in deep neural networks. The proposed model is trained to simultaneously minimize the sum of supervised and unsupervised cost functions by backpropagation, avoiding the need for layer-wise pre-training. Our work builds on top of the Ladder network proposed by Valpola [1] which we extend by combining the model with supervision. We show th...
متن کاملVirtual Adversarial Ladder Networks For Semi-supervised Learning
Semi-supervised learning (SSL) partially circumvents the high cost of labelling data by augmenting a small labeled dataset with a large and relatively cheap unlabeled dataset drawn from the same distribution. This paper offers a novel interpretation of two deep learning-based SSL approaches, ladder networks and virtual adversarial training (VAT), as applying distributional smoothing to their re...
متن کاملRecurrent Ladder Networks
We propose a recurrent extension of the Ladder networks [22] whose structure is motivated by the inference required in hierarchical latent variable models. We demonstrate that the recurrent Ladder is able to handle a wide variety of complex learning tasks that benefit from iterative inference and temporal modeling. The architecture shows close-to-optimal results on temporal modeling of video da...
متن کاملAdversarial Ladder Networks
The use of unsupervised data in addition to supervised data has lead to a significant improvement when training discriminative neural networks. However, the best results were achieved with a training process that is divided in two parts: first an unsupervised pre-training step is done for initializing the weights of the network and after these weights are refined with the use of supervised data...
متن کامل